8 research outputs found

    ZZ into 4â„“ expected sensitivity with the first ATLAS data

    Get PDF
    The qq->ZZ->4l process is of great interest because it has not been studied in the c.m. energies that the LHC will reach, it is the irreducible background to Higgs->ZZ->4l searches and it can be used to probe physics beyond the SM through the measurement of Triple Gauge Coupling parameters. In this talk, the analysis method for the measurement of the cross-section and neutral TGCs with simulated ATLAS data, and the expectations for early data taking will be presented

    Distributed and parallel sparse convex optimization for radio interferometry with PURIFY

    Full text link
    Next generation radio interferometric telescopes are entering an era of big data with extremely large data sets. While these telescopes can observe the sky in higher sensitivity and resolution than before, computational challenges in image reconstruction need to be overcome to realize the potential of forthcoming telescopes. New methods in sparse image reconstruction and convex optimization techniques (cf. compressive sensing) have shown to produce higher fidelity reconstructions of simulations and real observations than traditional methods. This article presents distributed and parallel algorithms and implementations to perform sparse image reconstruction, with significant practical considerations that are important for implementing these algorithms for Big Data. We benchmark the algorithms presented, showing that they are considerably faster than their serial equivalents. We then pre-sample gridding kernels to scale the distributed algorithms to larger data sizes, showing application times for 1 Gb to 2.4 Tb data sets over 25 to 100 nodes for up to 50 billion visibilities, and find that the run-times for the distributed algorithms range from 100 milliseconds to 3 minutes per iteration. This work presents an important step in working towards computationally scalable and efficient algorithms and implementations that are needed to image observations of both extended and compact sources from next generation radio interferometers such as the SKA. The algorithms are implemented in the latest versions of the SOPT (https://github.com/astro-informatics/sopt) and PURIFY (https://github.com/astro-informatics/purify) software packages {(Versions 3.1.0)}, which have been released alongside of this article.Comment: 25 pages, 5 figure

    Large-scale benchmarks of the Time-Warp/Graph-Theoretical Kinetic Monte Carlo approach for distributed on-lattice simulations of catalytic kinetics

    Get PDF
    We extend the work of Ravipati et al.[Comput. Phys. Commun., 2022, 270, 108148] in benchmarking the performance of large-scale, distributed, on-lattice kinetic Monte Carlo (KMC) simulations. Our software package, Zacros, employs a graph-theoretical approach to KMC, coupled with the Time-Warp algorithm for parallel discrete event simulations. The lattice is divided into equal subdomains, each assigned to a single processor; the cornerstone of the Time-Warp algorithm is the state queue, to which snapshots of the KMC (lattice) state are saved regularly, enabling historical KMC information to be corrected when conflicts occur at the subdomain boundaries. Focusing on three model systems, we highlight the key Time-Warp parameters that can be tuned to optimise KMC performance. The frequency of state saving, controlled by the state saving interval, δsnap, is shown to have the largest effect on performance, which favours balancing the overhead of re-simulating KMC history with that of writing state snapshots to memory. Also important is the global virtual time (GVT) computation interval, ΔτGVT, which has little direct effect on the progress of the simulation but controls how often the state queue memory can be freed up. We find that a vector data structure is, in general, more favourable than a linked list for storing the state queue, due to the reduced time required for allocating and de-allocating memory. These findings will guide users in maximising the efficiency of Zacros or other distributed KMC software, which is a vital step towards realising accurate, meso-scale simulations of heterogeneous catalysis

    Accelerating LHC event generation with simplified pilot runs and fast PDFs

    Get PDF
    Poor computing efficiency of precision event generators for LHC physics has become a bottleneck for Monte-Carlo event simulation campaigns. We provide solutions to this problem by focusing on two major components of general-purpose event generators: The PDF evaluator and the matrix-element generator. For a typical production setup in the ATLAS experiment, we show that the two can consume about 80% of the total runtime. Using NLO simulations of pp→ℓ+ℓ-+jets and pp→tt¯+jets as an example, we demonstrate that the computing footprint of Lhapdf and Sherpa can be reduced by factors of order 10, while maintaining the formal accuracy of the event sample. The improved codes are made publicly available

    Principles for automated and reproducible benchmarking

    Get PDF
    The diversity in processor technology used by High Performance Computing (HPC) facilities is growing, and so applications must be written in such a way that they can attain high levels of performance across a range of different CPUs, GPUs, and other accelerators. Measuring application performance across this wide range of platforms becomes crucial, but there are significant challenges to do this rigorously, in a time efficient way, whilst assuring results are scientifically meaningful, reproducible, and actionable. This paper presents a methodology for measuring and analysing the performance portability of a parallel application and shares a software framework which combines and extends adopted technologies to provide a usable benchmarking tool. We demonstrate the flexibility and effectiveness of the methodology and benchmarking framework by showcasing a variety of benchmarking case studies which utilise a stable of supercomputing resources at a national scale

    Distributed and parallel sparse convex optimization for radio interferometry with PURIFY

    No full text
    Next generation radio interferometric telescopes are entering an era of big data with extremely large data sets. While these telescopes can observe the sky in higher sensitivity and resolution than before, computational challenges in image reconstruction need to be overcome to realize the potential of forthcoming telescopes. New methods in sparse image reconstruction and convex optimization techniques (cf. compressive sensing) have shown to produce higher fidelity reconstructions of simulations and real observations than traditional methods. This article presents distributed and parallel algorithms and implementations to perform sparse image reconstruction, with significant practical considerations that are important for implementing these algorithms for Big Data. We benchmark the algorithms presented, showing that they are considerably faster than their serial equivalents. We then pre-sample gridding kernels to scale the distributed algorithms to larger data sizes, showing application times for 1 Gb to 2.4 Tb data sets over 25 to 100 nodes for up to 50 billion visibilities, and find that the run-times for the distributed algorithms range from 100 milliseconds to 3 minutes per iteration. This work presents an important step in working towards computationally scalable and efficient algorithms and implementations that are needed to image observations of both extended and compact sources from next generation radio interferometers such as the SKA. The algorithms are implemented in the latest versions of the SOPT (https://github.com/astro-informatics/sopt) and PURIFY (https://github.com/astro-informatics/purify) software packages {(Versions 3.1.0)}, which have been released alongside of this article

    Accelerating LHC event generation with simplified pilot runs and fast PDFs

    Get PDF
    Poor computing efficiency of precision event generators for LHC physics has become a bottleneck for Monte-Carlo event simulation campaigns. We provide solutions to this problem by focusing on two major components of general-purpose event generators: The PDF evaluator and the matrix-element generator. For a typical production setup in the ATLAS experiment, we show that the two can consume about 80% of the total runtime. Using NLO simulations of pp→ℓ+ℓ−+jets and pp→tt¯+jets as an example, we demonstrate that the computing footprint of LHAPDF and SHERPA can be reduced by factors of order 10, while maintaining the formal accuracy of the event sample. The improved codes are made publicly available
    corecore